Considering Misconceptions in Automatic Essay Scoring with A-TEST - Amrita Test Evaluation and Scoring Tool
نویسندگان
چکیده
In large classrooms with limited teacher time, there is a need for automatic evaluation of text answers and real-time personalized feedback during the learning process. In this paper, we discuss Amrita Test Evaluation & Scoring Tool (A-TEST), a text evaluation and scoring tool that learns from course materials and from human-rater scored text answers and also directly from teacher input. We use latent semantic analysis (LSA) to identify the key concepts. While most AES systems use LSA to compare students’ responses with a set of ideal essays, this ignores learning the common misconceptions that students may have about a topic. A-TEST also uses LSA to learn misconceptions from the lowest scoring essays using this as a factor for scoring. ‘A-TEST’ was evaluated using two datasets of 1400 and 1800 pre-scored text answers that were manually scored by two teachers. The scoring accuracy and kappa scores between the derived ‘A-TEST’ model and the human raters were comparable to those between the human raters.
منابع مشابه
An Evaluation of IntelliMetricTM Essay Scoring System Using Responses to GMAT® AWA Prompts
The Graduate Management Admission Council® (GMAC®) has long benefited from advances in automated essay scoring. When GMAC® adopted ETS® e-rater® in 1999, the Council’s flagship product, the Graduate Management Admission Test® (GMAT®), became the first large-scale assessment to incorporate automated essay scoring. The change was controversial at the time (Iowa State Daily, 1999; Calfee, 2000). T...
متن کاملPuzzle test: A tool for non-analytical clinical reasoning assessment
Most contemporary clinical reasoning tests typically assess non-automatic thinking. Therefore, a test is needed to measure automatic reasoning or pattern recognition, which has been largely neglected in clinical reasoning tests. The Puzzle Test (PT) is dedicated to assess automatic clinical reasoning in routine situations. This test has been introduced first in 2009 by Monajemi et al in the Oly...
متن کاملComputerized Evaluation of Essays
A " hot topic " in educational measurement is the area of computerized essay test evaluation. As essays have been incorporated into many of the standardized testing programs, and computerized testing is being implemented in more and more instances, there are many instances of where computers have evaluated the essay portion of standardized tests. In an environment where humans typically evaluat...
متن کاملAutomatic scoring of short handwritten essays in reading comprehension tests
Reading comprehension is largely tested in schools using handwritten responses. The paper describes computational methods of scoring such responses using handwriting recognition and automatic essay scoring technologies. The goal is to assign to each handwritten response a score which is comparable to that of a human scorer even though machine handwriting recognition methods have high transcript...
متن کاملThe Role of Centering Theory's Rough-Shift in the Teaching and Evaluation of Writing Skills
Existing software systems for automated essay scoring can provide NLP researchers with opportunities to test certain theoretical hypotheses, including some derived from Centering Theory. In this study we employ ETS's e-rater essay scoring system to examine whether local discourse coherence, as de ned by a measure of Rough-Shift transitions, might be a signi cant contributor to the evaluation of...
متن کامل